Discrete-time survival forests with Hellinger distance decision trees
نویسندگان
چکیده
منابع مشابه
Increasing Skew Insensitivity of Decision Trees with Hellinger Distance
Learning from unbalanced datasets presents a convoluted problem in which traditional learning algorithms typically perform poorly. The heuristics used in learning tend to favor the larger, less important classes in such problems. While other methods, like sampling, have been introduced to combat imbalance, these tend to be computationally expensive. This paper proposes Hellinger distance as a m...
متن کاملDiscrete-Time Survival Trees and Forests with Time-Varying Covariates: Application to Bankruptcy Data
Discrete-time survival data with time-varying covariates are often encountered in practice. One such example is bankruptcy studies where the status of each firm is available on a yearly basis. Moreover, these studies often use financial and accounting based ratios to predict bankruptcy. These ratios are also yearly measures and hence are time-varying. In this paper, we propose a new survival tr...
متن کاملDecision Forests with Oblique Decision Trees
Ensemble learning schemes have shown impressive increases in prediction accuracy over single model schemes. We introduce a new decision forest learning scheme, whose base learners are Minimum Message Length (MML) oblique decision trees. Unlike other tree inference algorithms,MMLoblique decision tree learning does not over-grow the inferred trees. The resultant trees thus tend to be shallow and ...
متن کاملHellinger distance
In this lecture, we will introduce a new notion of distance between probability distributions called Hellinger distance. Using some of the nice properties of this distance, we will generalize the fooling set argument for deterministic protocols to the randomized setting. We will then use this to prove a Ω(n) lower bound for the communication complexity of Disjointness. We will also see how this...
متن کاملForests of Decision Trees
In many cases it is better to extract a set of decision trees and a set of possible logical data descriptions instead of a single model. Methods for creating forests of decision trees based on Separability of Split Value (SSV) criterion are presented. Preliminary results confirm their usefulness in understanding data structures.
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Data Mining and Knowledge Discovery
سال: 2020
ISSN: 1384-5810,1573-756X
DOI: 10.1007/s10618-020-00682-z